- Home
- Search Results
- Page 1 of 1
Search for: All records
-
Total Resources4
- Resource Type
-
0003000001000000
- More
- Availability
-
40
- Author / Contributor
- Filter by Author / Creator
-
-
Wang, Linnan (4)
-
Bosilca, George (2)
-
Guo, Tian (2)
-
Wu, Wei (2)
-
Zhao, Yiyang (2)
-
Dongarra, Jack (1)
-
Fonseca, Rodrigo (1)
-
Herlihy, Maurice (1)
-
Liu, Hang (1)
-
Luo, Xi (1)
-
Patinyasakdikul, Thananon (1)
-
Tian, Yuandong (1)
-
Yang, Kevin (1)
-
Zhang, Junyu (1)
-
Zhang, Tianjun (1)
-
#Tyler Phillips, Kenneth E. (0)
-
#Willis, Ciara (0)
-
& Abreu-Ramos, E. D. (0)
-
& Abramson, C. I. (0)
-
& Abreu-Ramos, E. D. (0)
-
- Filter by Editor
-
-
& Spizer, S. M. (0)
-
& . Spizer, S. (0)
-
& Ahn, J. (0)
-
& Bateiha, S. (0)
-
& Bosch, N. (0)
-
& Brennan K. (0)
-
& Brennan, K. (0)
-
& Chen, B. (0)
-
& Chen, Bodong (0)
-
& Drown, S. (0)
-
& Ferretti, F. (0)
-
& Higgins, A. (0)
-
& J. Peters (0)
-
& Kali, Y. (0)
-
& Ruiz-Arias, P.M. (0)
-
& S. Spitzer (0)
-
& Sahin. I. (0)
-
& Spitzer, S. (0)
-
& Spitzer, S.M. (0)
-
(submitted - in Review for IEEE ICASSP-2024) (0)
-
-
Have feedback or suggestions for a way to improve these results?
!
Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher.
Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?
Some links on this page may take you to non-federal websites. Their policies may differ from this site.
-
Zhao, Yiyang; Wang, Linnan; Yang, Kevin; Zhang, Tianjun; Guo, Tian; Tian, Yuandong (, International Conference on Learning Representations (ICLR'22))
-
Luo, Xi; Wu, Wei; Bosilca, George; Patinyasakdikul, Thananon; Wang, Linnan; Dongarra, Jack (, The 27th International Symposium on High-Performance Parallel and Distributed Computing)The increase in scale and heterogeneity of high-performance computing (HPC) systems predispose the performance of Message Passing Interface (MPI) collective communications to be susceptible to noise, and to adapt to a complex mix of hardware capabilities. The designs of state of the art MPI collectives heavily rely on synchronizations; these designs magnify noise across the participating processes, resulting in significant performance slowdown. Therefore, such design philosophy must be reconsidered to efficiently and robustly run on the large-scale heterogeneous platforms. In this paper, we present ADAPT, a new collective communication framework in Open MPI, using event-driven techniques to morph collective algorithms to heterogeneous environments. The core concept of ADAPT is to relax synchronizations, while mamtaining the minimal data dependencies of MPI collectives. To fully exploit the different bandwidths of data movement lanes in heterogeneous systems, we extend the ADAPT collective framework with a topology-aware communication tree. This removes the boundaries of different hardware topologies while maximizing the speed of data movements. We evaluate our framework with two popular collective operations: broadcast and reduce on both CPU and GPU clusters. Our results demonstrate drastic performance improvements and a strong resistance against noise compared to other state of the art MPI libraries. In particular, we demonstrate at least 1.3X and 1.5X speedup for CPU data and 2X and 10X speedup for GPU data using ADAPT event-based broadcast and reduce operations.more » « less
-
Wang, Linnan; Wu, Wei; Zhang, Junyu; Liu, Hang; Bosilca, George; Herlihy, Maurice; Fonseca, Rodrigo (, HPDC '20: Proceedings of the 29th International Symposium on High-Performance Parallel and Distributed Computing)
An official website of the United States government

Full Text Available